Adversarial Delays in Online Strongly-Convex Optimization

نویسندگان

  • Daniel Khashabi
  • Kent Quanrud
  • Amirhossein Taghvaei
چکیده

We consider the problem of strongly-convex online optimization in presence of adversarial delays [1]; in a T -iteration online game, the feedback of the player’s query at time t is arbitrarily delayed by an adversary for dt rounds and delivered before the game ends, at iteration t+ dt − 1. Specifically for online-gradient-descent algorithm we show it has a simple regret bound of O (∑T t=1 log(1 + dt t ) ) . This gives a clear and simple bound without resorting any distributional and limiting assumptions on the delays. We further show how this result encompasses and generalizes several of the existing known results in the literature. Specifically it matches the celebrated logarithmic regret O (log T ) [2] when there are no delays (i.e. dt = 1) and the regret bound of O (τ log T ) [3] for constant delays dt = τ .

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

High Probability Bounds for Bandit Stochastic and Adversarial Strongly Convex Optimization

In this note, we consider the problem of multiple point bandit online convex optimization introduced in [2]. The focus is on the setting where the loss functions are strongly convex aiming at proving high probability bounds for multiple point bandit online strongly convex optimization. We consider two scenarios: the stochastic setup where all the received training examples are i.i.d samples, an...

متن کامل

Stochastic and Adversarial Online Learning without Hyperparameters

Most online optimization algorithms focus on one of two things: performing well in adversarial settings by adapting to unknown data parameters (such as Lipschitz constants), typically achieving O( √ T ) regret, or performing well in stochastic settings where they can leverage some structure in the losses (such as strong convexity), typically achieving O(log(T )) regret. Algorithms that focus on...

متن کامل

No-Regret Learnability for Piecewise Linear Losses

In the convex optimization approach to online regret minimization, many methodshave been developed to guarantee a O(√T ) bound on regret for subdifferentiableconvex loss functions with bounded subgradients, by using a reduction to linearloss functions. This suggests that linear loss functions tend to be the hardest onesto learn against, regardless of the underlying d...

متن کامل

Online Learning with Adversarial Delays

We study the performance of standard online learning algorithms when the feedback is delayed by an adversary. We show that online-gradient-descent [1] and follow-the-perturbed-leader [2] achieve regret O( √ D) in the delayed setting, where D is the sum of delays of each round’s feedback. This bound collapses to an optimal O( √ T ) bound in the usual setting of no delays (where D = T ). Our main...

متن کامل

Applications of strong convexity--strong smoothness duality to learning with matrices

It is known that a function is strongly convex with respect to some norm if and only if its conjugate function is strongly smooth with respect to the dual norm. This result has already been found to be a key component in deriving and analyzing several learning algorithms. Utilizing this duality, we isolate a single inequality which seamlessly implies both generalization bounds and online regret...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1605.06201  شماره 

صفحات  -

تاریخ انتشار 2016